Lower bounds on minimax rates for nonparametric regression with additive sparsity and smoothness

نویسندگان

  • Garvesh Raskutti
  • Martin J. Wainwright
  • Bin Yu
چکیده

We study minimax rates for estimating high-dimensional nonparametric regression models with sparse additive structure and smoothness constraints. More precisely, our goal is to estimate a function f∗ : R → R that has an additive decomposition of the form f(X1, . . . ,Xp) = ∑ j∈S h ∗ j (Xj), where each component function h ∗ j lies in some class H of “smooth” functions, and S ⊂ {1, . . . , p} is an unknown subset with cardinality s = |S|. Given n i.i.d. observations of f∗(X) corrupted with additive white Gaussian noise where the covariate vectors (X1,X2,X3, ...,Xp) are drawn with i.i.d. components from some distribution P, we determine lower bounds on the minimax rate for estimating the regression function with respect to squared-L(P) error. Our main result is a lower bound on the minimax rate that scales as max ( s log(p/s) n , s ǫ 2 n(H) ) . The first term reflects the sample size required for performing subset selection, and is independent of the function class H. The second term s ǫn(H) is an s-dimensional estimation term corresponding to the sample size required for estimating a sum of s univariate functions, each chosen from the function class H. It depends linearly on the sparsity index s but is independent of the global dimension p. As a special case, if H corresponds to functions that are m-times differentiable (an m-order Sobolev space), then the s-dimensional estimation term takes the form sǫn(H) ≍ s n−2m/(2m+1). Either of the two terms may be dominant in different regimes, depending on the relation between the sparsity and smoothness of the additive decomposition.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Minimax-optimal rates for high-dimensional sparse additive models over kernel classes

Sparse additive models are families of d-variate functions that have the additive decomposition f = ∑ j∈S f ∗ j , where S is an unknown subset of cardinality s ≪ d. We consider the case where each component function f j lies in a reproducing kernel Hilbert space, and analyze an l1 kernel-based method for estimating the unknown function f . Working within a highdimensional framework that allows ...

متن کامل

Minimax-Optimal Rates For Sparse Additive Models Over Kernel Classes Via Convex Programming

Sparse additive models are families of d-variate functions that have the additive decomposition f∗ = ∑ j∈S f ∗ j , where S is a unknown subset of cardinality s d. We consider the case where each component function f∗ j lies in a reproducing kernel Hilbert space, and analyze a simple kernel-based convex program for estimating the unknown function f∗. Working within a high-dimensional framework t...

متن کامل

Multivariate Dyadic Regression Trees for Sparse Learning Problems

We propose a new nonparametric learning method based on multivariate dyadic regression trees (MDRTs). Unlike traditional dyadic decision trees (DDTs) or classification and regression trees (CARTs), MDRTs are constructed using penalized empirical risk minimization with a novel sparsity-inducing penalty. Theoretically, we show that MDRTs can simultaneously adapt to the unknown sparsity and smooth...

متن کامل

Sparse additive regression on a regular lattice

We consider estimation in a sparse additive regression model with the design points on a regular lattice. We establish the minimax convergence rates over Sobolev classes and propose a Fourier-based rate optimal estimator which is adaptive to the unknown sparsity and smoothness of the response function. The estimator is derived within a Bayesian formalism but can be naturally viewed as a penaliz...

متن کامل

Minimax Lower Bounds for Noisy Matrix Completion Under Sparse Factor Models

This paper examines fundamental error characteristics for a general class of matrix completion problems, where matrix of interest is a product of two a priori unknown matrices, one of which is sparse, and the observations are noisy. Our main contributions come in the form of minimax lower bounds for the expected per-element squared error for these problems under several noise/corruption models;...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2009